Goto

Collaborating Authors

 cycle time


LMC: Large Model Collaboration with Cross-assessment for Training-Free Open-Set Object Recognition (Supplementary Material)

Neural Information Processing Systems

In Figure 1, we compare our LMC framework with the baseline Softmax, and present qualitative results on the TinyImageNet dataset. Below, we discuss them in more detail. AUROC is a widely-used threshold-independent evaluation metric. Both authors contributed equally to the work. Before entering the inference process, similar to our framework, Softmax also pre-stores certain CLIP and DINO features to make the inference process more efficient.



Throughput-OptimalTopology Design forCross-SiloFederatedLearning

Neural Information Processing Systems

Federated learning (FL) "involves training statistical models over remote devices or siloed data centers,suchasmobile phones orhospitals, whilekeepingdatalocalized"[56]because ofprivacy concerns orlimitedcommunication resources. Hence, clients only communicate with apotentially far-away (e.g., in another continent) orchestrator and do not Recent experimental and theoretical work suggests that, in practice,the first effect has been over-estimated by classic worst-caseconvergencebounds.


A Cross-Embodiment Gripper Benchmark for Rigid-Object Manipulation in Aerial and Industrial Robotics

Vagas, Marek, Varga, Martin, Romancik, Jaroslav, Majercak, Ondrej, Suarez, Alejandro, Ollero, Anibal, Vanderborght, Bram, Virgala, Ivan

arXiv.org Artificial Intelligence

Abstract--Robotic grippers are increasingly deployed across industrial, collaborative, and aerial platforms, where each embodiment imposes distinct mechanical, energetic, and operational constraints. Established YCB and NIST benchmarks quantify grasp success, force, or timing on a single platform, but do not evaluate cross-embodiment transferability or energy-aware performance, capabilities essential for modern mobile and aerial manipulation. This letter introduces the Cross-Embodiment Gripper Benchmark (CEGB), a compact and reproducible benchmarking suite extending YCB and selected NIST metrics with three additional components: a transfer-time benchmark measuring the practical effort required to exchange embodiments, an energy-consumption benchmark evaluating grasping and holding efficiency, and an intent-specific ideal payload assessment reflecting design-dependent operational capability. T ogether, these metrics characterize both grasp performance and the suitability of reusing a single gripper across heterogeneous robotic systems. A lightweight self-locking gripper prototype is implemented as a reference case. Experiments demonstrate rapid embodiment transfer (median 17.6 s across user groups), low holding energy for gripper prototype ( 1.5 J per 10 s), and consistent grasp performance with cycle times of 3.2-3.9 CEGB thus provides a reproducible foundation for cross-platform, energy-aware evaluation of grippers in aerial and manipulators domains. Robotic grasping has been extensively investigated across industrial, collaborative, and aerial domains.



Learning to Optimize Capacity Planning in Semiconductor Manufacturing

Andelfinger, Philipp, Bi, Jieyi, Zhu, Qiuyu, Zhou, Jianan, Zhang, Bo, Zhang, Fei Fei, Chan, Chew Wye, Gan, Boon Ping, Cai, Wentong, Zhang, Jie

arXiv.org Artificial Intelligence

In manufacturing, capacity planning is the process of allocating production resources in accordance with variable demand. The current industry practice in semiconductor manufacturing typically applies heuristic rules to prioritize actions, such as future change lists that account for incoming machine and recipe dedications. However, while offering interpretability, heuristics cannot easily account for the complex interactions along the process flow that can gradually lead to the formation of bottlenecks. Here, we present a neural network-based model for capacity planning on the level of individual machines, trained using deep reinforcement learning. By representing the policy using a heterogeneous graph neural network, the model directly captures the diverse relationships among machines and processing steps, allowing for proactive decision-making. We describe several measures taken to achieve sufficient scalability to tackle the vast space of possible machine-level actions. Our evaluation results cover Intel's small-scale Minifab model and preliminary experiments using the popular SMT2020 testbed. In the largest tested scenario, our trained policy increases throughput and decreases cycle time by about 1.8% each.


A Rollout-Based Algorithm and Reward Function for Resource Allocation in Business Processes

Middelhuis, Jeroen, Bukhsh, Zaharah, Adan, Ivo, Dijkman, Remco

arXiv.org Artificial Intelligence

Resource allocation plays a critical role in minimizing cycle time and improving the efficiency of business processes. Recently, Deep Reinforcement Learning (DRL) has emerged as a powerful technique to optimize resource allocation policies in business processes. In the DRL framework, an agent learns a policy through interaction with the environment, guided solely by reward signals that indicate the quality of its decisions. However, existing algorithms are not suitable for dynamic environments such as business processes. Furthermore, existing DRL-based methods rely on engineered reward functions that approximate the desired objective, but a misalignment between reward and objective can lead to undesired decisions or suboptimal policies. To address these issues, we propose a rollout-based DRL algorithm and a reward function to optimize the objective directly. Our algorithm iteratively improves the policy by evaluating execution trajectories following different actions. Our reward function directly decomposes the objective function of minimizing the cycle time, such that trial-and-error reward engineering becomes unnecessary. We evaluated our method in six scenarios, for which the optimal policy can be computed, and on a set of increasingly complex, realistically sized process models. The results show that our algorithm can learn the optimal policy for the scenarios and outperform or match the best heuristics on the realistically sized business processes.


References [1 ]

Neural Information Processing Systems

Mahmoud Assran et al. "Stochastic Gradient Push for Distributed Deep Learning". Keith Bonawitz et al. "Practical secure aggregation for privacy-preserving machine learning". Pierre Courtiol et al. "Deep learning-based classification of mesothelioma improves prediction "Distributed nonconvex optimization over time-varying networks". "Dual Averaging for Distributed Optimization: Matt Fredrikson, Somesh Jha, and Thomas Ristenpart. "Model inversion attacks that exploit Transactions on The Built Environment 37 (1998). Zhanhong Jiang et al. "Collaborative deep learning in fixed topology networks". Can Karakus et al. "Straggler Mitigation in Distributed Optimization Through Data Encoding". "Federated Optimization:Distributed Optimization Beyond the Datacenter". Jakub Konecný et al. "Federated Optimization: Distributed Machine Learning for On-Device Songze Li et al. "Near-Optimal Straggler Mitigation for Distributed Gradient Methods".